Skip to content

feat(deepinfra): add GLM-4.7-Flash model#781

Open
Dagnan wants to merge 1 commit intoanomalyco:devfrom
Dagnan:add-glm-4.7-flash-deepinfra
Open

feat(deepinfra): add GLM-4.7-Flash model#781
Dagnan wants to merge 1 commit intoanomalyco:devfrom
Dagnan:add-glm-4.7-flash-deepinfra

Conversation

@Dagnan
Copy link

@Dagnan Dagnan commented Feb 2, 2026

Summary

Add zai-org/GLM-4.7-Flash model configuration to the DeepInfra provider. GLM-4.7-Flash is a 30B-A3B MoE model released by Z.AI (Zhipu AI) in January 2026, positioned as the "free-ish" variant of the GLM-4.7 family with strong performance across coding, reasoning, and generative tasks.

Changes

  • New file: providers/deepinfra/models/zai-org/GLM-4.7-Flash.toml
  • Model ID: zai-org/GLM-4.7-Flash
  • Family: glm-flash
  • Pricing: $0.06 input / $0.40 output per 1M tokens
  • Context window: 202,752 tokens
  • Output limit: 16,384 tokens (DeepInfra enforced limit)
  • Features: reasoning, tool calling, temperature control
  • Open weights: true

References

Note: The release date uses 2026-01-19 as this is the logical correct date (Z.AI's website shows a typo with 2025, but GLM-4.7 was released in December 2025, making January 2026 the correct timeline for the Flash variant).

Add zai-org/GLM-4.7-Flash to DeepInfra provider

Ultraworked with [Sisyphus](https://github.com/code-yeongyu/oh-my-opencode)

Co-authored-by: Sisyphus <clio-agent@sisyphuslabs.ai>
@Dagnan
Copy link
Author

Dagnan commented Feb 5, 2026

Hi, is there an issue with this PR? I see the validation hasn't run.

@rekram1-node
Copy link
Contributor

idk why it cancelled, weird

tool_call = true
open_weights = true

[cost]
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

isn't this an interleaved reasoning model?

Shouldn't the reasoning_content setting be here

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants